Goto

Collaborating Authors

 deep reinforcement learning and search


Review for NeurIPS paper: Combining Deep Reinforcement Learning and Search for Imperfect-Information Games

Neural Information Processing Systems

The paper was reviewed by experts on the topic and discussed after authors rebuttal. Results were found to be interesting and valuable. The reviewers comments should be taken into account while preparing the final version of the paper.


Review for NeurIPS paper: Combining Deep Reinforcement Learning and Search for Imperfect-Information Games

Neural Information Processing Systems

According to the papre, ReBeL is a novel method to deal with two-player zero-sum imperfect-information games. And it may be able to be used to solve other inperfect-information problem. And the domain of this paper, multi-agents RL in imperfect-information, has high relevance to NIPS. 6. The experiment compared the module with human player, which is a strong evidence of the exploitability of ReBeL.


Combining Deep Reinforcement Learning and Search for Imperfect-Information Games

Neural Information Processing Systems

The combination of deep reinforcement learning and search at both training and test time is a powerful paradigm that has led to a number of successes in single-agent settings and perfect-information games, best exemplified by AlphaZero. However, prior algorithms of this form cannot cope with imperfect-information games. This paper presents ReBeL, a general framework for self-play reinforcement learning and search that provably converges to a Nash equilibrium in any two-player zero-sum game. In the simpler setting of perfect-information games, ReBeL reduces to an algorithm similar to AlphaZero. Results in two different imperfect-information games show ReBeL converges to an approximate Nash equilibrium. We also show ReBeL achieves superhuman performance in heads-up no-limit Texas hold'em poker, while using far less domain knowledge than any prior poker AI.